Specified complexity

Specified complexity is an argument proposed by William Dembski and used by him and others to promote intelligent design. According to Dembski, the concept is intended to formalize a property that singles out patterns that are both specified and complex. Dembski states that specified complexity is a reliable marker of design by an intelligent agent, a central tenet to intelligent design which Dembski argues for in opposition to modern evolutionary theory. The concept of specified complexity is widely regarded as mathematically unsound and has not been the basis for further independent work in information theory, the theory of complex systems, or biology.[1][2][3] Specified complexity is one of the two main arguments used by intelligent design proponents, the other being irreducible complexity.

In Dembski's terminology, a specified pattern is one that admits short descriptions, whereas a complex pattern is one that is unlikely to occur by chance. Dembski argues that it is impossible for specified complexity to exist in patterns displayed by configurations formed by unguided processes. Therefore, Dembski argues, the fact that specified complex patterns can be found in living things indicates some kind of guidance in their formation, which is indicative of intelligence. Dembski further argues that one can rigorously show by applying no free lunch theorems the inability of evolutionary algorithms to select or generate configurations of high specified complexity.

In intelligent design literature, an intelligent agent is one that chooses between different possibilities and has, by supernatural means and methods, caused life to arise.[4] Specified complexity is what Dembski terms an "explanatory filter" which can recognize design by detecting complex specified information (CSI). The filter is based on the premise that the categories of regularity, chance, and design are, according to Dembski, "mutually exclusive and exhaustive." Complex specified information detects design because it detects what characterizes intelligent agency; it detects the actualization of one among many competing possibilities.

A study by Wesley Elsberry and Jeffrey Shallit states that "Dembski's work is riddled with inconsistencies, equivocation, flawed use of mathematics, poor scholarship, and misrepresentation of others' results".[5] Another objection concerns Dembski's calculation of probabilities. According to Martin Nowak, a Harvard professor of mathematics and evolutionary biology "We cannot calculate the probability that an eye came about. We don't have the information to make the calculation".[6] Critics also reject applying specified complexity to infer design as an argument from ignorance.

Contents

Definition

Orgel's original use

The term "specified complexity" was originally coined by origin of life researcher Leslie Orgel to denote what distinguishes living things from non-living things:

In brief, living organisms are distinguished by their specified complexity. Crystals are usually taken as the prototypes of simple well-specified structures, because they consist of a very large number of identical molecules packed together in a uniform way. Lumps of granite or random mixtures of polymers are examples of structures that are complex but not specified. The crystals fail to qualify as living because they lack complexity; the mixtures of polymers fail to qualify because they lack specificity.[7]

The term was later employed by physicist Paul Davies in a similar manner:

Living organisms are mysterious not for their complexity per se, but for their tightly specified complexity[8]

Dembski's definition

For Dembski, specified complexity is a property which can be observed in living things. However, whereas Orgel used the term for biological features which are considered in science to have arisen through a process of evolution, Dembski says that it describes features which cannot form through "undirected" evolution—and concludes that it allows one to infer intelligent design. While Orgel employed the concept in a qualitative way, Dembski's use is intended to be quantitative. Dembski's use of the concept dates to his 1998 monograph The Design Inference. Specified complexity is fundamental to his approach to intelligent design, and each of his subsequent books has also dealt significantly with the concept. He has stated that, in his opinion, "if there is a way to detect design, specified complexity is it."[9]

Dembski asserts that specified complexity is present in a configuration when it can be described by a pattern that displays a large amount of independently specified information and is also complex, which he defines as having a low probability of occurrence. He provides the following examples to demonstrate the concept: "A single letter of the alphabet is specified without being complex. A long sentence of random letters is complex without being specified. A Shakespearean sonnet is both complex and specified." [10]

In his earlier papers Dembski defined complex specified information (CSI) as being present in a specified event whose probability did not exceed 1 in 10150, which he calls the universal probability bound. In that context, "specified" meant what in later work he called "pre-specified", that is specified before any information about the outcome is known. The value of the universal probability bound corresponds to the inverse of the upper limit of "the total number of [possible] specified events throughout cosmic history," as calculated by Dembski.[11] Anything below this bound has CSI. The terms "specified complexity" and "complex specified information" are used interchangeably. In more recent papers Dembski has redefined the universal probability bound, with reference to another number, corresponding to the total number of bit operations that could possibly have been performed in the entire history of the universe.

Dembski asserts that CSI exists in numerous features of living things, such as DNA and other functional biological molecules, and argues that it cannot be generated by the only known natural mechanisms of physical law and chance, or by their combination. He argues that this is so because laws can only shift around or lose information, but do not produce it, and chance can produce complex unspecified information, or simple specified information, but not CSI; he provides a mathematical analysis that he claims demonstrates that law and chance working together cannot generate CSI, either. Moreover, he claims that CSI is holistic, with the whole being greater than the sum of the parts, and that this decisively eliminates Darwinian evolution as a possible means of its creation. Dembski maintains that by process of elimination, CSI is best explained as being due to intelligence, and is therefore a reliable indicator of design.

Law of conservation of information

Dembski formulates and proposes a law of conservation of information as follows:

This strong proscriptive claim, that natural causes can only transmit CSI but never originate it, I call the Law of Conservation of Information. Immediate corollaries of the proposed law are the following:

  1. The specified complexity in a closed system of natural causes remains constant or decreases.
  2. The specified complexity cannot be generated spontaneously, originate endogenously or organize itself (as these terms are used in origins-of-life research).
  3. The specified complexity in a closed system of natural causes either has been in the system eternally or was at some point added exogenously (implying that the system, though now closed, was not always closed).
  4. In particular any closed system of natural causes that is also of finite duration received whatever specified complexity it contains before it became a closed system.[12]

Dembski notes that the term "Law of Conservation of Information" was previously used by Peter Medawar in his book The Limits of Science (1984) "to describe the weaker claim that deterministic laws cannot produce novel information."[13] The actual validity and utility of Dembski's proposed law are uncertain; it is neither widely used by the scientific community nor cited in mainstream scientific literature. A 2002 essay by Erik Tellgren provided a mathematical rebuttal of Dembski's law and concludes that it is "mathematically unsubstantiated." [14]

Specificity

In a more recent paper,[15] Dembski provides an account which he claims is simpler and adheres more closely to the theory of statistical hypothesis testing as formulated by Ronald Fisher. In general terms, Dembski proposes to view design inference as a statistical test to reject a chance hypothesis P on a space of outcomes Ω.

Dembski's proposed test is based on the Kolmogorov complexity of a pattern T that is exhibited by an event E that has occurred. Mathematically, E is a subset of Ω, the pattern T specifies a set of outcomes in Ω and E is a subset of T. Quoting Dembski[16]

Thus, the event E might be a die toss that lands six and T might be the composite event consisting of all die tosses that land on an even face.

Kolmogorov complexity provides a measure of the computational resources needed to specify a pattern (such as a DNA sequence or a sequence of alphabetic characters).[17] Given a pattern T, the number of other patterns may have Kolmogorov complexity no larger than that of T is denoted by φ(T). The number φ(T) thus provides a ranking of patterns from the simplest to the most complex. For example, for a pattern T which describes the bacterial flagellum, Dembski claims to obtain the upper bound φ(T) ≤ 1020.

Dembski defines specified complexity of the pattern T under the chance hypothesis P as

 \sigma= - \log_2 [R \times \varphi(T) \times \operatorname{P}(T)],

where P(T) is the probability of observing the pattern T, R is the number of "replicational resources" available "to witnessing agents". R corresponds roughly to repeated attempts to create and discern a pattern. Dembski then asserts that R can be bounded by 10120. This number is supposedly justified by a result of Seth Lloyd[18] in which he determines that the number of elementary logic operations that can have performed in the universe over its entire history cannot exceed 10120 operations on 1090 bits.

Dembski's main claim is that the following test can be used to infer design for a configuration: There is a target pattern T that applies to the configuration and whose specified complexity exceeds 1. This condition can be restated as the inequality

 10^{120} \times \varphi(T) \times \operatorname{P}(T) < \frac{1}{2}.

Dembski's explanation of specified complexity

Dembski's expression σ is unrelated to any known concept in information theory, though he claims he can justify its relevance as follows: An intelligent agent S witnesses an event E and assigns it to some reference class of events Ω and within this reference class considers it as satisfying a specification T. Now consider the quantity φ(T) × P(T) (where P is the "chance" hypothesis):

Think of S as trying to determine whether an archer, who has just shot an arrow at a large wall, happened to hit a tiny target on that wall by chance. The arrow, let us say, is indeed sticking squarely in this tiny target. The problem, however, is that there are lots of other tiny targets on the wall. Once all those other targets are factored in, is it still unlikely that the archer could have hit any of them by chance?

In addition, we need to factor in what I call the replicational resources associated with T, that is, all the opportunities to bring about an event of T's descriptive complexity and improbability by multiple agents witnessing multiple events.

According to Dembski, the number of such "replicational resources" can be bounded by "the maximal number of bit operations that the known, observable universe could have performed throughout its entire multi-billion year history", which according to Lloyd is 10120.

However, according to Elsberry and Shallit, "[specified complexity] has not been defined formally in any reputable peer-reviewed mathematical journal, nor (to the best of our knowledge) adopted by any researcher in information theory."[19]

Calculation of specified complexity

Thus far, Dembski's only attempt at calculating the specified complexity of a naturally occurring biological structure is in his book No Free Lunch, for the bacterial flagellum of E. coli. This structure can be described by the pattern "bidirectional rotary motor-driven propeller". Dembski estimates that there are at most 1020 patterns described by four basic concepts or fewer, and so his test for design will apply if

 \operatorname{P}(T) < \frac{1}{2} \times 10^{-140}.

However, Dembski says that the precise calculation of the relevant probability "has yet to be done", although he also claims that some methods for calculating these probabilities "are now in place".

These methods assume that all of the constituent parts of the flagellum must have been generated completely at random, a scenario that biologists do not seriously consider. He justifies this approach by appealing to Michael Behe's concept of "irreducible complexity" (IC), which leads him to assume that the flagellum could not come about by any gradual or step-wise process. The validity of Dembski's particular calculation is thus wholly dependent on Behe's IC concept, and therefore susceptible to its criticisms, of which there are many.

To arrive at the ranking upper bound of 1020 patterns, Dembski considers a specification pattern for the flagellum defined by the (natural language) predicate "bidirectional rotary motor-driven propeller", which he regards as being determined by four independently chosen basic concepts. He furthermore assumes that English has the capability to express at most 105 basic concepts (an upper bound on the size of a dictionary). Dembski then claims that we can obtain the rough upper bound of

 10^{20}= 10^5 \times 10^5 \times 10^5 \times 10^5

for the set of patterns described by four basic concepts or fewer.

From the standpoint of Kolmogorov complexity theory, this calculation is problematic. Quoting Ellsberry and Shallit[20] "Natural language specification without restriction, as Dembski tacitly permits, seems problematic. For one thing, it results in the Berry paradox". These authors add: "We have no objection to natural language specifications per se, provided there is some evident way to translate them to Dembski's formal framework. But what, precisely, is the space of events Ω here?"

Criticisms

The soundness of Dembski's concept of specified complexity and the validity of arguments based on this concept are widely disputed. A frequent criticism (see Elsberry and Shallit) is that Dembski has used the terms "complexity", "information" and "improbability" interchangeably. These numbers measure properties of things of different types: Complexity measures how hard it is to describe an object (such as a bitstring), information measures how close to uniform a random probability distribution is and improbability measures how unlikely an event is given a probability distribution.

When Dembski's mathematical claims on specific complexity are interpreted to make them meaningful and conform to minimal standards of mathematical usage, they usually turn out to be false. Dembski often sidesteps these criticisms by responding that he is not "in the business of offering a strict mathematical proof for the inability of material mechanisms to generate specified complexity".[21] Yet on page 150 of No Free Lunch he claims he can prove his thesis mathematically: "In this section I will present an in-principle mathematical argument for why natural causes are incapable of generating complex specified information." Others have pointed out that a crucial calculation on page 297 of No Free Lunch is off by a factor of approximately 1065.[22]

Dembski's calculations show how a simple smooth function cannot gain information. He therefore concludes that there must be a designer to obtain CSI. However, natural selection has a branching mapping from one to many (replication) followed by pruning mapping of the many back down to a few (selection). When information is replicated, some copies can be differently modified while others remain the same, allowing information to increase. These increasing and reductional mappings were not modeled by Dembski. In other words, Dembski's calculations do not model birth and death. This basic flaw in his modeling renders all of Dembski's subsequent calculations and reasoning in No Free Lunch irrelevant because his basic model does not reflect reality. Since the basis of No Free Lunch relies on this flawed argument, the entire thesis of the book collapses.[23]

According to Martin Nowak, a Harvard professor of mathematics and evolutionary biology "We cannot calculate the probability that an eye came about. We don't have the information to make the calculation".[6]

Dembski's critics note that specified complexity, as originally defined by Leslie Orgel, is precisely what Darwinian evolution is supposed to create. Critics maintain that Dembski uses "complex" as most people would use "absurdly improbable". They also claim that his argument is a tautology: CSI cannot occur naturally because Dembski has defined it thus. They argue that to successfully demonstrate the existence of CSI, it would be necessary to show that some biological feature undoubtedly has an extremely low probability of occurring by any natural means whatsoever, something which Dembski and others have almost never attempted to do. Such calculations depend on the accurate assessment of numerous contributing probabilities, the determination of which is often necessarily subjective. Hence, CSI can at most provide a "very high probability", but not absolute certainty.

Another criticism refers to the problem of "arbitrary but specific outcomes". For example, if a coin is tossed randomly 1000 times, the probability of any particular outcome occurring is roughly one in 10300. For any particular specific outcome of the coin-tossing process, the a priori probability that this pattern occurred is thus one in 10300, which is astronomically smaller than Dembski's universal probability bound of one in 10150. Yet we know that the post hoc probability of its happening is exactly one, since we observed it happening. This is similar to the observation that it is unlikely that any given person will win a lottery, but, eventually, a lottery will have a winner; to argue that it is very unlikely that any one player would win is not the same as proving that there is the same chance that no one will win. Similarly, it has been argued that "a space of possibilities is merely being explored, and we, as pattern-seeking animals, are merely imposing patterns, and therefore targets, after the fact."[12]

Apart from such theoretical considerations, critics cite reports of evidence of the kind of evolutionary "spontanteous generation" that Dembski claims is too improbable to occur naturally. For example, in 1982, B.G. Hall published research demonstrating that after removing a gene that allows sugar digestion in certain bacteria, those bacteria, when grown in media rich in sugar, rapidly evolve new sugar-digesting enzymes to replace those removed.[24] Another widely cited example is the discovery of nylon eating bacteria that produce enzymes only useful for digesting synthetic materials that did not exist prior to the invention of nylon in 1935.

Other commentators have noted that evolution through selection is frequently used to design certain electronic, aeronautic and automotive systems which are considered problems too complex for human "intelligent designers".[25] This strongly contradicts the argument that an intelligent designer is required for the most complex systems. Such evolutionary techniques can lead to designs that are difficult to understand or evaluate since no human understands which trade-offs were made in the evolutionary process, something which mimics our poor understanding of biological systems.

Dembski's book No Free Lunch was criticised for not addressing the work of researchers who use computer simulations to investigate artificial life. According to Jeffrey Shallit:

The field of artificial life evidently poses a significant challenge to Dembski's claims about the failure of evolutionary algorithms to generate complexity. Indeed, artificial life researchers regularly find their simulations of evolution producing the sorts of novelties and increased complexity that Dembski claims are impossible.[22]

See also

Notes and references

  1. ^ Rich Baldwin (2005). "Information Theory and Creationism: William Dembski". TalkOrigins Archive. http://www.talkorigins.org/faqs/information/dembski.html. Retrieved 2010-05-10. 
  2. ^ Mark Perakh, (2005). Dembski "displaces Darwinism" mathematically -- or does he?
  3. ^ Jason Rosenhouse, (2001). How Anti-Evolutionists Abuse Mathematics The Mathematical Intelligencer, Vol. 23, No. 4, Fall 2001, pp. 3-8.
  4. ^ "no intelligent agent who is strictly physical could have presided over the origin of the universe or the origin of life." - Dembski The Act of Creation: Bridging Transcendence and Immanence Dembski has also said that "intelligent design should be understood as the evidence that God has placed in nature to show that the physical world is the product of intelligence and not simply the result of mindless material forces" Why President Bush Got It Right about Intelligent Design William A. Dembski DesignInference.com, August 4, 2005. Dembski has also said that "Intelligent design is just the Logos theology of John's Gospel restated in the idiom of information theory." Signs of Intelligence A Primer on the Discernment of Intelligent Design. William A. Dembski. Touchstone Journal, Volume 12, Issue 4, July/August 1999.
  5. ^ Wesley Elsberry and Jeffrey Shallit, (2003). Information Theory, Evolutionary Computation, and Dembski’s “Complex Specified Information”
  6. ^ a b Martin Nowak (2005). Time Magazine, 15 August 2005, page 32
  7. ^ Leslie Orgel (1973). The Origins of Life, p. 189.
  8. ^ Paul Davies (1999). The Fifth Miracle p. 112
  9. ^ William A. Dembski (2002). No Free Lunch, p. 19.
  10. ^ William A. Dembski (1999). Intelligent Design, p. 47.
  11. ^ William A. Dembski (2004). The Design Revolution: Answering the Toughest Questions About Intelligent Design, p. 85.
  12. ^ a b William A. Dembski (1998) Intelligent Design as a Theory of Information.
  13. ^ "Searching Large Spaces: Displacement and the No Free Lunch Regress (356k PDF)", pp. 15-16, describing an argument made by Michael Shermer in How We Believe: Science, Skepticism, and the Search for God, 2nd ed. (2003).
  14. ^ On Dembski's law of conservation of information Erik Tellgren. talkreason.org, 2002. (PDF file)
  15. ^ William A. Dembski (2005). Specification: The Pattern that Signifies intelligence
  16. ^ (loc. cit. p 16)
  17. ^ Michael Sipser (1997). Introduction to the Theory of Computation, PWS Publishing Company.
  18. ^ Seth Lloyd (2002), Computational capacity of the universe, Phys. Rev. Lett. 88(23):790 1- 4. See also arXiv:quant/ph0110141.
  19. ^ Elsberry and Shallit, (2003)p 14
  20. ^ Elsberry and Shallit, (2003)p 12
  21. ^ William A. Dembski, (2002). If Only Darwinists Scrutinized Their Own Work as Closely: A Response to "Erik".
  22. ^ a b Jeffrey Shallit (2002) A review of Dembski's No Free Lunch
  23. ^ Thomas D. Schneider. (2002) Dissecting Dembski's "Complex Specified Information"
  24. ^ B.G. Hall (1982). "Evolution of a regulated operon in the laboratory", Genetics, 101(3-4):335-44. In PubMed.
  25. ^ Evolutionary algorithms now surpass human designers New Scientist, 28 July 2007

References

External links